# Ultra-low-bit quantization
Qwen3 30B A3B GGUF
Apache-2.0
Qwen3-30B-A3B is a large language model based on Qwen3-30B-A3B-Base, supporting text generation tasks, optimized for memory efficiency with ultra-low-bit quantization technology.
Large Language Model
Q
Mungert
2,135
1
Qwen3 14B GGUF
Apache-2.0
Qwen3-14B is a GGUF format model generated from Qwen/Qwen3-14B-Base, supporting text generation tasks and optimized for memory efficiency using IQ-DynamicGate ultra-low-bit quantization technology.
Large Language Model
Q
Mungert
1,597
6
GLM Z1 9B 0414 GGUF
MIT
GLM-Z1-9B-0414 is a bilingual text generation model supporting both Chinese and English, utilizing the GGUF format and suitable for various quantization levels, from BF16 to ultra-low-bit quantization (1-2 bits).
Large Language Model Supports Multiple Languages
G
Mungert
1,598
3
Qwen2.5 3B Instruct GGUF
Other
Ultra-low-bit quantization (1-2 bit) model using IQ-DynamicGate technology, suitable for memory-constrained devices and efficient inference scenarios
Large Language Model English
Q
Mungert
704
4
Llama 3.2 3B Instruct GGUF
Llama-3.2-3B-Instruct GGUF is a 3B-parameter large language model released by Meta, utilizing IQ-DynamicGate technology for ultra-low-bit quantization (1-2 bits), optimizing inference performance while maintaining memory efficiency.
Large Language Model Supports Multiple Languages
L
Mungert
656
3
Granite 3.3 8b Instruct GGUF
Apache-2.0
Ultra-low-bit quantization (1-2 bits) language model using IQ-DynamicGate technology, suitable for memory-constrained environments
Large Language Model
G
Mungert
759
2
Deepcoder 14B Preview GGUF
MIT
Ultra-low-bit quantization (1-2 bits) model using IQ-DynamicGate technology, suitable for memory-constrained devices and edge computing scenarios
Large Language Model English
D
Mungert
1,764
6
Qwq 32B GGUF
Apache-2.0
Ultra-low-bit quantization (1-2 bit) large language model using IQ-DynamicGate technology, supporting multilingual text generation tasks
Large Language Model English
Q
Mungert
5,770
17
Llama 3.1 70B Instruct GGUF
An ultra-low-bit (1-2 bit) quantized model based on Llama-3.1-70B, utilizing IQ-DynamicGate technology for adaptive precision quantization, enhancing accuracy while maintaining memory efficiency.
Large Language Model Supports Multiple Languages
L
Mungert
19.52k
3
Olympiccoder 32B GGUF
Apache-2.0
OlympicCoder-32B is a code generation model based on Qwen2.5-Coder-32B-Instruct, employing IQ-DynamicGate ultra-low-bit quantization technology for efficient inference in memory-constrained environments.
Large Language Model English
O
Mungert
361
3
Llama 3 3 Nemotron Super 49B V1 GGUF
Other
A 49B-parameter large language model utilizing IQ-DynamicGate ultra-low-bit quantization technology, supporting 1-2 bit precision-adaptive quantization, optimized for memory efficiency and inference speed
Large Language Model English
L
Mungert
434
4
Gemma 3 27b It GGUF
GGUF quantized version of Gemma 3 with 27B parameters, supporting image-text interaction tasks
Text-to-Image
G
Mungert
4,034
6
EXAONE Deep 32B GGUF
Other
EXAONE-Deep-32B is a 32B-parameter large language model supporting English and Korean, specifically designed for text generation tasks.
Large Language Model Supports Multiple Languages
E
Mungert
2,249
3
EXAONE Deep 7.8B GGUF
Other
A 7.8B-parameter model featuring ultra-low-bit quantization (1-2 bits) using IQ-DynamicGate technology, supporting English and Korean text generation tasks.
Large Language Model Supports Multiple Languages
E
Mungert
1,791
5
Granite 3.2 8b Instruct GGUF
Apache-2.0
IBM Granite series 8B parameter instruction-tuned language model, employing IQ-DynamicGate ultra-low-bit quantization technology, suitable for efficient inference scenarios
Large Language Model
G
Mungert
1,048
2
Qwen2.5 7B Instruct 1M GGUF
Apache-2.0
Qwen2.5-7B-Instruct-1M is an instruction-tuned version based on Qwen2.5-7B, employing IQ-DynamicGate ultra-low-bit quantization (1-2 bits), suitable for efficient inference in memory-constrained environments.
Large Language Model English
Q
Mungert
1,342
4
Qwen2.5 14B Instruct 1M GGUF
Apache-2.0
Qwen2.5-14B-Instruct-1M is an instruction-tuned model based on Qwen2.5-14B, supporting text generation tasks and suitable for chat scenarios.
Large Language Model English
Q
Mungert
1,600
3
Meta Llama 3 8B Instruct GGUF
An IQ-DynamicGate ultra-low-bit quantization (1-2 bit) model based on Llama-3-8B-Instruct, utilizing precision-adaptive quantization technology to enhance inference accuracy while maintaining extreme memory efficiency.
Large Language Model English
M
Mungert
1,343
3
Llama 3.1 8B Instruct GGUF
Llama-3.1-8B-Instruct is an instruction-tuned version based on Llama-3-8B, utilizing IQ-DynamicGate technology for ultra-low-bit quantization (1-2 bits), enhancing accuracy while maintaining memory efficiency.
Large Language Model Supports Multiple Languages
L
Mungert
1,073
3
Featured Recommended AI Models